Feature Scaling for Kernel Fisher Discriminant Analysis Using Leave-One-Out Cross Validation
نویسندگان
چکیده
Kernel fisher discriminant analysis (KFD) is a successful approach to classification. It is well known that the key challenge in KFD lies in the selection of free parameters such as kernel parameters and regularization parameters. Here we focus on the feature-scaling kernel where each feature individually associates with a scaling factor. A novel algorithm, named FS-KFD, is developed to tune the scaling factors and regularization parameters for the feature-scaling kernel. The proposed algorithm is based on optimizing the smooth leave-one-out error via a gradient-descent method and has been demonstrated to be computationally feasible. FS-KFD is motivated by the following two fundamental facts: the leave-one-out error of KFD can be expressed in closed form and the step function can be approximated by a sigmoid function. Empirical comparisons on artificial and benchmark data sets suggest that FS-KFD improves KFD in terms of classification accuracy.
منابع مشابه
E cient leave-one-out cross-validation of kernel Fisher discriminant classi'ers
Mika et al. (in: Neural Network for Signal Processing, Vol. IX, IEEE Press, New York, 1999; pp. 41–48) apply the “kernel trick” to obtain a non-linear variant of Fisher’s linear discriminant analysis method, demonstrating state-of-the-art performance on a range of benchmark data sets. We show that leave-one-out cross-validation of kernel Fisher discriminant classi'ers can be implemented with a ...
متن کاملOptimally regularised kernel Fisher discriminant classification
Mika, Rätsch, Weston, Schölkopf and Müller [Mika, S., Rätsch, G., Weston, J., Schölkopf, B., & Müller, K.-R. (1999). Fisher discriminant analysis with kernels. In Neural networks for signal processing: Vol. IX (pp. 41-48). New York: IEEE Press] introduce a non-linear formulation of Fisher's linear discriminant, based on the now familiar "kernel trick", demonstrating state-of-the-art performance...
متن کاملEfficient leave-one-out cross-validation of kernel fisher discriminant classifiers
Mika et al. [1] apply the “kernel trick” to obtain a non-linear variant of Fisher’s linear discriminant analysis method, demonstrating state-of-the-art performance on a range of benchmark datasets. We show that leave-one-out cross-validation of kernel Fisher discriminant classifiers can be implemented with a computational complexity of only O(l3) operations rather than the O(l4) of a näıve impl...
متن کاملEfficient cross-validation of kernel fisher discriminant classifiers
Mika et al. [1] introduce a non-linear formulation of the Fisher discriminant based the well-known “kernel trick”, later shown to be equivalent to the Least-Squares Support Vector Machine [2, 3]. In this paper, we show that the cross-validation error can be computed very efficiently for this class of kernel machine, specifically that leave-one-out cross-validation can be performed with a comput...
متن کاملClassification of Aurora Series Image Based on Eigenvalue-scaling Kernel Fisher Discriminant Analysis and Klt
Aurora is the typical ionosphere track generated by the interaction of solar wind and magnetosphere. This paper proposed a method based on eigenvalue-scaling kernel fisher discriminant analysis and Karhunen-Loeve Transform (KLT) to take advantage of interband correlation between aurora images to detect the change of aurora in serial time. Conventional classification algorithms are incapable of ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Neural computation
دوره 18 4 شماره
صفحات -
تاریخ انتشار 2006